Classification Methods with Reject Option Based on Convex Risk Minimization

نویسندگان

  • Ming Yuan
  • Marten H. Wegkamp
چکیده

In this paper, we investigate the problem of binary classification with a reject option in which one can withhold the decision of classifying an observation at a cost lower than that of misclassification. Since the natural loss function is non-convex so that empirical risk minimization easily becomes infeasible, the paper proposes minimizing convex risks based on surrogate convex loss functions. A necessary and sufficient condition for infinite sample consistency (both risks share the same minimizer) is provided. Moreover, we show that the excess risk can be bounded through the excess surrogate risk under appropriate conditions. These bounds can be tightened by a generalized margin condition. The impact of the results is illustrated on several commonly used surrogate loss functions.

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Lasso type classifiers with a reject option

This paper discusses structural risk minimization in the setting of classification with a reject option. Binary classification is about classifying observations that take values in an arbitrary feature space X into one of two classes, labelled −1 or +1. A discriminant function f : X → R yields a classifier sgn(f(x)) ∈ {−1,+1} that represents our guess of the label Y of a future observation X an...

متن کامل

Double Ramp Loss Based Reject Option Classifier

The performance of a reject option classifiers is quantified using 0 − d − 1 loss where d ∈ (0, .5) is the loss for rejection. In this paper, we propose double ramp loss function which gives a continuous upper bound for (0 − d − 1) loss. Our approach is based on minimizing regularized risk under the double ramp loss using difference of convex programming. We show the effectiveness of our approa...

متن کامل

Consistency of Multiclass Empirical Risk Minimization Methods Based on Convex Loss

The consistency of classification algorithm plays a central role in statistical learning theory. A consistent algorithm guarantees us that taking more samples essentially suffices to roughly reconstruct the unknown distribution. We consider the consistency of ERM scheme over classes of combinations of very simple rules (base classifiers) in multiclass classification. Our approach is, under some...

متن کامل

A Projected Alternating Least square Approach for Computation of Nonnegative Matrix Factorization

Nonnegative matrix factorization (NMF) is a common method in data mining that have been used in different applications as a dimension reduction, classification or clustering method. Methods in alternating least square (ALS) approach usually used to solve this non-convex minimization problem.  At each step of ALS algorithms two convex least square problems should be solved, which causes high com...

متن کامل

Efficiency of Classification Methods Based on Empirical Risk Minimization

EFFICIENCY OF CLASSIFICATION METHODS BASED ON EMPIRICAL RISK MINIMIZATION V. I. Norkin a and M. A. Keyzer b UDC 519:234:24:85 A binary classification problem is reduced to the minimization of convex regularized empirical risk functionals in a reproducing kernel Hilbert space. The solution is searched for in the form of a finite linear combination of kernel support functions (Vapnik’s support ve...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

عنوان ژورنال:
  • Journal of Machine Learning Research

دوره 11  شماره 

صفحات  -

تاریخ انتشار 2010